Latent Predictor Networks for Code Generation

نویسندگان

  • Wang Ling
  • Phil Blunsom
  • Edward Grefenstette
  • Karl Moritz Hermann
  • Tomás Kociský
  • Fumin Wang
  • Andrew Senior
چکیده

Many language generation tasks require the production of text conditioned on both structured and unstructured inputs. We present a novel neural network architecture which generates an output sequence conditioned on an arbitrary number of input functions. Crucially, our approach allows both the choice of conditioning context and the granularity of generation, for example characters or tokens, to be marginalised, thus permitting scalable and effective training. Using this framework, we address the problem of generating programming code from a mixed natural language and structured specification. We create two new data sets for this paradigm derived from the collectible trading card games Magic the Gathering and Hearthstone. On these, and a third preexisting corpus, we demonstrate that marginalising multiple predictors allows our model to outperform strong benchmarks.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Optimal Combined and Adaptive Protection of Active Distribution Networks Considering Different System Topologies Incorporating Optimal Selection of Standard Relay Curves

The change in the topology of active distribution networks (ADNs) is one of the essential challenges that might affect the protection schemes. The conventional protection schemes based on base topology result in some coordination constraint violations in other topologies due to the outage of upstream substations and distributed generation units. In this article, new combinational and adaptive p...

متن کامل

SyncGAN: Synchronize the Latent Space of Cross-modal Generative Adversarial Networks

Generative adversarial network (GAN) has achieved impressive success on cross-domain generation, but it faces difficulty in cross-modal generation due to the lack of a common distribution between heterogeneous data. Most existing methods of conditional based cross-modal GANs adopt the strategy of one-directional transfer and have achieved preliminary success on text-to-image transfer. Instead o...

متن کامل

Latent Constraints: Learning to Generate Conditionally from Unconditional Generative Models

Deep generative neural networks have proven effective at both conditional and unconditional modeling of complex data distributions. Conditional generation enables interactive control, but creating new controls often requires expensive retraining. In this paper, we develop a method to condition generation without retraining the model. By post-hoc learning latent constraints, value functions that...

متن کامل

Analysis and Comparison of Load Flow Methods for Distribution Networks Considering Distributed Generation

Conventional passive distribution networks are changing to modern active distribution networks which are not radial. Conventional load flow methods should be modified for new distribution networks analysis. In modern distribution networks distributed generation (DG) units are embedded with conventional and/or renewable resources. DG units are generally modeled as PV or PQ nodes which inject ...

متن کامل

Compact and Efficient Presentation Conversion Routines

Presentation conversion is a key operation in any development environment for distributed applications, such as Corba, Java-RMI, DCE or ASN.1-based environments. It is also well-known performance bottleneck in high-speed network communication. Presentation conversion code is usually generated by an automatic code generation tool referred to as stub compiler. The quality of the code generated by...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • CoRR

دوره abs/1603.06744  شماره 

صفحات  -

تاریخ انتشار 2016